Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DRAFT] add harm categories to AdvBench Dataset #732

Draft
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

paulinek13
Copy link
Contributor

Description

This PR aims to resolve #730 by adding a way to manually assign harm categories to the AdvBench dataset and enabling filtering support based on those categories.

Marked as a draft PR since I'm seeking confirmation on the approach

Tests and Documentation

Copy link
Contributor Author

@paulinek13 paulinek13 Mar 4, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I actually tried using the categories from the AdvBench paper (profanity, graphic depictions, threatening behavior, misinformation, discrimination, cybercrime, and dangerous or illegal suggestions) but, honestly, the results aren't great: pretty much everything ends up under dangerous or illegal suggestions, which isn't super useful IMO 😅

{
    "0": ["cybercrime", "dangerous or illegal suggestions"],
    "1": ["cybercrime", "dangerous or illegal suggestions"],
    "2": ["dangerous or illegal suggestions"],
    "3": ["dangerous or illegal suggestions"],
    "4": ["cybercrime", "dangerous or illegal suggestions"],
    "5": ["dangerous or illegal suggestions"],
    "6": ["discrimination", "dangerous or illegal suggestions"],
    "7": ["dangerous or illegal suggestions"],
	...
}

So then I tested out that Collaborative, Human-Centered Taxonomy of AI, Algorithmic, and Automation Harms instead. It breaks things down into main categories (e.g.: Physical, Financial, Psychological) and subcategories (e.g.: Privacy Loss, Economic Instability, Coercion/manipulation), and honestly, the output looks a lot more meaningful. I think it captures nuance better instead of labelling everything into one or two types.
For now, I only categorized the first 50 prompts with the help of the Claude 3.7 model and I put the results in this file pyrit/datasets/harm_categories/adv_bench_dataset.json.

Well... I really hope I’m not making things more chaotic 😅 I just want to make sure we’re getting the best possible results!
Let me know what you think :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

BUG harm categories for AdvBench Dataset aren't added yet
2 participants